homotopy theory, (∞,1)-category theory, homotopy type theory
flavors: stable, equivariant, rational, p-adic, proper, geometric, cohesive, directed…
models: topological, simplicial, localic, …
see also algebraic topology
Introductions
Definitions
Paths and cylinders
Homotopy groups
Basic facts
Theorems
The Gram–Schmidt process is an algorithm which takes as input an ordered linear basis of an inner product space and produces as output an ordered orthonormal basis.
In terms of matrices, the Gram–Schmidt process is a procedure of factorization of a invertible matrix in the general linear group (or ) as a product where
is an upper triangular matrix
is an orthogonal (or unitary) matrix.
As such, this provides matrix decomposition in specialization of the more general Iwasawa decomposition for (connected) semisimple Lie groups. In the case of real matrices, decomposition into an orthogonal and upper triangular matrix is often denoted and called QR decomposition or QR factorization.
Since the factorization depends smoothly on the parameters, the Gram–Schmidt procedure enables the reduction of the structure group of an inner product vector bundle (e.g., the tangent bundle of a Riemannian manifold or a Kähler manifold) from to the orthogonal group (or the unitary group ).
In this section, “basis” is understood to signify an ordered independent set whose linear span is dense in a Hilbert space seen as a metric space. We will describe the Gram–Schmidt process as applied to a -dimensional Hilbert space for some cardinal with a basis consisting of vectors.
The orthonormal basis produced as output is defined recursively by a) subtracting the orthogonal projection to the closed subspace generated by all previous vectors and b) normalizing. We denote the orthogonal projection onto a closed subspace by and the normalization of a vector by . For ordinals define
where is the closure of the span of , noting that the projection is known to exist, since is complete. This can be rewritten more explicitly using transfinite recursion as
where the sum on the right is well defined by the Bessel inequality, i.e. only countably many coefficients are non-zero and they are square-summable. A simple (transfinite) inductive argument shows that the are unit vectors orthogonal to each other, and that the span of is equal to the span of for . Therefore is an orthonormal basis of .
(Application to non-bases)
If we apply the Gram–Schmidt process to a well-ordered independent set whose closed linear span is not all of , we still get an orthonormal basis of the subspace . If we apply the Gram–Schmidt process to a dependent set, then we will eventually run into a vector whose norm is zero, so we will not be able to take . In that case, however, we can simply remove from the set and continue; then we will still get an orthonormal basis of the closed linear span. (This conclusion is not generally valid in constructive mathematics, since it relies on excluded middle applied to the statement that . However, it does work to discrete fields, such as the algebraic closure of the rationals, as seen in elementary undergraduate linear algebra.)
There is an alternative algorithm via Gaussian elimination which for a set of vectors produces an orthogonal basis for its linear span (Pursell & Trimble 1991).
In the special case that the original vectors are linearly independent and after normalizing the resulting orthogonal basis to an orthonormal basis, the output of this algorithm is an orthonormal basis as produced also by the Gram-Schmidt process.
(LU-decomposition of positive semidefinite matrices)
Every positive semidefinite matrix has a matrix product-decomposition
where
is
is an upper triangular matrix.
(Pursell & Trimble 1991, p. 4)
(orthogonal basis of linear span via LU-decomposition)
Let be a tuple of vectors of the same length, hence let
be the matrix whose th column is .
Since the matrix product of with its transpose matrix is necessarily a positive semidefinite symmetric matrix it has an LU-decomposition according to Lemma :
Then the column vectors of the matrix
constitute an orthogonalization of the original tuple of vectors in that the non-zero form an orthogonal linear basis of the linear span of the .
(Pursell & Trimble 1991, top of p. 5)
That the matrix in Prop. is lower triangular and invertible (by Lemma ) means that its inverse matrix is also a lower triangular matrix which reflects a process of row reduction from to .
Accordingly, the orthogonalization in Prop. may be understood as applying Gauss elimination to
A classic illustration of Gram–Schmidt is the production of the Legendre polynomials.
Let be the Hilbert space of square integrable functions in the closed interval, equipped with the standard inner product defined by
By the Stone-Weierstrass theorem, the space of polynomials is dense in according to its standard inclusion, and so the polynomials form an ordered basis of .
Applying the Gram–Schmidt process as above, one readily computes the first few orthonormal functions:
The classical Legendre polynomials are scalar multiplies of the functions , adjusted so that ; they satisfy the orthogonality relations
where is the Kronecker delta.
Many aspects of the Gram–Schmidt process can be categorified so as to apply to 2-Hilbert spaces as indicated at Schur's lemma in the section In terms of categorical algebra;
We will illustrate the basic idea with an example that was suggested to us by James Dolan. (See also at permutation representation the sections Examples – Virtual permutation representations.)
Consider the category of representations over the complex numbers of the symmetric group . (As a running example, we consider ; up to isomorphism, there are five irreducible representations
classified by the five Young diagrams of size 4. To save space, we denote these as , , , , .)
The irreducible representations of form a -orthonormal basis in the sense that any two of them satisfy the relation
(where denotes the hom vector space in , and indicates a direct sum of copies of ). In fact, the irreducible representations are uniquely determined up to isomorphism by these relations.
There is however another way of associating representations to partitions or Young diagrams. Namely, consider the subgroup of permutations which take each row of a Young diagram or Young tableau of size to itself; this forms a parabolic subgroup of , conjugate to one of type where is the length of the row of the Young diagram. The group acts transitively on the orbit space of cosets
and these actions give linear permutation representations of . Equivalently, these are linear representations which are induced from the trivial representation along inclusions of parabolic subgroups.
We claim that
these representations form a -linear basis of the representation ring ,
we may calculate their characters using a categorified Gram–Schmidt process.
We indicate the proof:
Given two such parabolic subgroups , in , the -inner product
may be identified with the free vector space on the set of double cosets .
One may count the number of double cosets by hand in a simple case like . That is, for the 5 representations induced from the 5 parabolic subgroups corresponding to the 5 Young diagrams listed above, the dimensions of the 2-inner products are the sizes of the corresponding double coset spaces . These numbers form a matrix as follows (following the order of the partitions listed above):
To reiterate: this matrix is the decategorification (a matrix of dimensions) of a matrix of -inner products where the -entry is of the form
where the are induced from inclusions of parabolic subgroups. The are -linear combinations of irreducible representations which form a -orthonormal basis, and we may perform a series of elementary row operations which convert this matrix into an upper triangular matrix, and which will turn out to be the decategorified form of the 2-matrix with entries
where is the irrep corresponding to the th Young diagram (as listed above). The upper triangular matrix is
and we read off from the columns the following decompositions into irreducible components:
The last representation is the regular representation of (because the parabolic subgroup is trivial). Since we know from general theory that the multiplicity of the irreducible in the regular representation is its dimension, we get as a by-product the dimensions of the from the expression for :
(the first of the is the trivial representation, and the last is the alternating representation).
The row operations themselves can be assembled as the lower triangular matrix
and from the rows we read off the irreducible representations as “virtual” (i.e., -linear) combinations of the parabolically induced representations :
which can be considered the result of the categorified Gram–Schmidt process.
It follows from these representations that the form a -linear basis of the representation ring .
Analogous statements hold for each symmetric group .
See also
The formulation of Gram-Schmidt via Gaussian elimination is due to
Evident generalization to the case of indefinite inner product spaces:
Last revised on June 17, 2024 at 13:55:29. See the history of this page for a list of all contributions to it.